Interpolation of stochastic grammar and word bigram models in natural language understanding
نویسندگان
چکیده
The paper shows the effects of combining a stochastic grammar with a word bigram language model by log-linear interpolation. It is divided into three main parts: The first part derives the stochastic grammar model and gives a sound theoretical motivation to incorporate word dependencies such as bigrams. The second part describes two different algorithmic approaches to the combination of both models by log-linear interpolation. The third part reports attribute error rate (AER) results measured on the Philips corpus of train time table inquiries that show a reduction of up to 9% relative. 1. STOCHASTIC MODEL OF NATURAL LANGUAGE UNDERSTANDING The Philips Natural Language Understanding (NLU) module is used in automated inquiry systems (AIS), such as train table enquiries [2], to analyze the word sequence of a user utterance. It does not try to find parse trees that cover the whole word sequence but breaks up the sequence into chunks, where each chunk belongs to a semantically meaningful concept. A stochastic context–free grammar is used to derive the word chunk from a concept. The chunking is useful since the spontaneous speech that occurs in dialogue applications is very ungrammatical. Thus, a robust NLU model concentrates on the useful parts of a user utterance. Other recent works also employ some kind of chunking, e.g. [6, 9]. The stochastic model of the Philips NLU module was developed by H. Aust in [1, p. 81]. Here, we show that this model can be derived from Bayes’ decision rule. This derivation gives a sound theoretical motivation to incorporate word dependencies such as bigrams. Bayes’ decision rule finds the most likely concept sequence K̂ = k̂1, . . . , k̂s, given the sequence O = o1, . . . , ot of acoustic observations. The derivation of the concept sequence K does not directly depend on the acoustic observations O but on a word sequence W = w1, . . . , wN derived from O as an intermediate result: K̂ = argmax K p(K|O)
منابع مشابه
Permugram language models
In natural languages, the words within an utterance are often correlated over large distances. Long-spanning con-textual eeects of this type cannot be eeciently and robustly captured by the traditional N-gram approaches of stochastic language modelling. We present a new kind of stochastic grammar | the permugram model. A permu-gram model is obtained by linear interpolation of a large number of ...
متن کاملExperience with a Stack Decoder-Based HMM CSR and Back-Off N-Gram Language Models
Stochastic language models are more useful than nonstochastic models because they contribute more information than a simple acceptance or rejection of a word sequence. Back-off N-gram language models [ I l l are an effective class of word based stochastic language model. The first part of this paper describes our experiences using the back-off language models in our time-synchronous decoder CSR...
متن کاملUsing a stochastic context-free grammar as a language model for speech recognition
This paper describes a number of experiments in adding new grammatical knowledge to the Berkeley Restaurant Project (BeRP), our medium-vocabulary (1300 word), speaker-independent, spontaneous continuous-speech understanding system (Jurafsky et al. 1994). We describe an algorithm for using a probabilistic Earley parser and a stochastic context-free grammar (SCFG) to generate word transition prob...
متن کاملParsing N-Best Lists of Handwritten Sentences
This paper investigates the application of a probabilistic parser for natural language on the list of the Nbest sentences produced by an off-line recognition system for cursive handwritten sentences. For the generation of the N-best sentence list an HMM-based recognizer including a bigram language model is used. The parsing of the sentences is achieved by a bottom-up chart parser for stochastic...
متن کاملA word graph interface for a flexible concept based speech understanding framework
In this paper, we introduce a word graph interface between speech and natural language processing systems within a flexible speech understanding framework based on stochastic concept modeling augmented with background ”filler” models. Each concept represents a set of phrases ( written as a context free grammar (CFG)) with the same meaning, and is compiled into a stochastic recursive transition ...
متن کامل